Toward better feature weighting algorithms: a focus on Relief
نویسندگان
چکیده
Feature weighting algorithms try to solve a problem of great importance nowadays in machine learning: The search of a relevance measure for the features of a given domain. This relevance is primarily used for feature selection as feature weighting can be seen as a generalization of it, but it is also useful to better understand a problem’s domain or to guide an inductor in its learning process. Relief family of algorithms are proven to be very effective in this task. Some other feature weighting methods are reviewed in order to give some context and then the different existing extensions to the original algorithm are explained. One of Relief’s known issues is the performance degradation of its estimates when redundant features are present. A novel theoretical definition of redundancy level is given in order to guide the work towards an extension of the algorithm that is more robust against redundancy. A new extension is presented that aims for improving the algorithms performance. Some experiments were driven to test this new extension against the existing ones with a set of artificial and real datasets and denoted that in certain cases it improves the weight’s estimation accuracy. 1 Overview Feature selection is undoubtedly one of the most important problems in machine learning, pattern recognition and information retrieval, among others. A feature selection algorithm is a computational solution that is motivated by a certain definition of relevance. However, the relevance of a feature may have several definitions depending on the objective that is looked after. The generic purpose pursued is the improvement of the inductive learner, either in terms of learning speed, generalization capacity or simplicity of the
منابع مشابه
Double Relief with progressive weighting function
Feature weighting algorithms try to solve a problem of great importance nowadays in machine learning: The search of a relevance measure for the features of a given domain. This relevance is primarily used for feature selection as feature weighting can be seen as a generalization of it, but it is also useful to better understand a problem’s domain or to guide an inductor in its learning process....
متن کاملFeature Weighting Method Based On Instance Correlation Using Discretization
In Machine Learning Process, several issues arise in identifying a suitable and quality set of features from which a classification model for a particular domain to be constructed. This paper addresses the problem of feature selection for machine learning through discretization approach. RELIEF is considered to be one of the most successful algorithms for assessing the quality of features. RELI...
متن کاملA Bi-objective Stochastic Optimization Model for Humanitarian Relief Chain by Using Evolutionary Algorithms
Due to the increasing amount of natural disasters such as earthquakes and floods and unnatural disasters such as war and terrorist attacks, Humanitarian Relief Chain (HRC) is taken into consideration of most countries. Besides, this paper aims to contribute humanitarian relief chains under uncertainty. In this paper, we address a humanitarian logistics network design problem including local dis...
متن کاملScalable extensions of the ReliefF algorithm for weighting and selecting features on the multi-label learning context
Multi-label learning has become an important area of research due to the increasing number of modern applications that contain multi-label data. The multi-label data are structured in a more complex way than single-label data. Consequently the development of techniques that allow the improvement in the performance of machine learning algorithms over multi-label data is desired. The feature weig...
متن کاملInvestigation of Term Weighting Schemes in Classification of Imbalanced Texts
Class imbalance problem in data, plays a critical role in use of machine learning methods for text classification since feature selection methods expect homogeneous distribution as well as machine learning methods. This study investigates two different kinds of feature selection metrics (one-sided and two-sided) as a global component of term weighting schemes (called as tffs) in scenarios where...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1509.03755 شماره
صفحات -
تاریخ انتشار 2015